452 research outputs found
Systems of Linear Equations over and Problems Parameterized Above Average
In the problem Max Lin, we are given a system of linear equations
with variables over in which each equation is assigned a
positive weight and we wish to find an assignment of values to the variables
that maximizes the excess, which is the total weight of satisfied equations
minus the total weight of falsified equations. Using an algebraic approach, we
obtain a lower bound for the maximum excess.
Max Lin Above Average (Max Lin AA) is a parameterized version of Max Lin
introduced by Mahajan et al. (Proc. IWPEC'06 and J. Comput. Syst. Sci. 75,
2009). In Max Lin AA all weights are integral and we are to decide whether the
maximum excess is at least , where is the parameter.
It is not hard to see that we may assume that no two equations in have
the same left-hand side and . Using our maximum excess results,
we prove that, under these assumptions, Max Lin AA is fixed-parameter tractable
for a wide special case: for an arbitrary fixed function
.
Max -Lin AA is a special case of Max Lin AA, where each equation has at
most variables. In Max Exact -SAT AA we are given a multiset of
clauses on variables such that each clause has variables and asked
whether there is a truth assignment to the variables that satisfies at
least clauses. Using our maximum excess results, we
prove that for each fixed , Max -Lin AA and Max Exact -SAT AA can
be solved in time This improves
-time algorithms for the two problems obtained by Gutin et
al. (IWPEC 2009) and Alon et al. (SODA 2010), respectively
A Hypergraph Dictatorship Test with Perfect Completeness
A hypergraph dictatorship test is first introduced by Samorodnitsky and
Trevisan and serves as a key component in their unique games based \PCP
construction. Such a test has oracle access to a collection of functions and
determines whether all the functions are the same dictatorship, or all their
low degree influences are Their test makes queries and has
amortized query complexity but has an inherent loss of
perfect completeness. In this paper we give an adaptive hypergraph dictatorship
test that achieves both perfect completeness and amortized query complexity
.Comment: Some minor correction
When Can Limited Randomness Be Used in Repeated Games?
The central result of classical game theory states that every finite normal
form game has a Nash equilibrium, provided that players are allowed to use
randomized (mixed) strategies. However, in practice, humans are known to be bad
at generating random-like sequences, and true random bits may be unavailable.
Even if the players have access to enough random bits for a single instance of
the game their randomness might be insufficient if the game is played many
times.
In this work, we ask whether randomness is necessary for equilibria to exist
in finitely repeated games. We show that for a large class of games containing
arbitrary two-player zero-sum games, approximate Nash equilibria of the
-stage repeated version of the game exist if and only if both players have
random bits. In contrast, we show that there exists a class of
games for which no equilibrium exists in pure strategies, yet the -stage
repeated version of the game has an exact Nash equilibrium in which each player
uses only a constant number of random bits.
When the players are assumed to be computationally bounded, if cryptographic
pseudorandom generators (or, equivalently, one-way functions) exist, then the
players can base their strategies on "random-like" sequences derived from only
a small number of truly random bits. We show that, in contrast, in repeated
two-player zero-sum games, if pseudorandom generators \emph{do not} exist, then
random bits remain necessary for equilibria to exist
A shortcut to (sun)flowers: Kernels in logarithmic space or linear time
We investigate whether kernelization results can be obtained if we restrict
kernelization algorithms to run in logarithmic space. This restriction for
kernelization is motivated by the question of what results are attainable for
preprocessing via simple and/or local reduction rules. We find kernelizations
for d-Hitting Set(k), d-Set Packing(k), Edge Dominating Set(k) and a number of
hitting and packing problems in graphs, each running in logspace. Additionally,
we return to the question of linear-time kernelization. For d-Hitting Set(k) a
linear-time kernelization was given by van Bevern [Algorithmica (2014)]. We
give a simpler procedure and save a large constant factor in the size bound.
Furthermore, we show that we can obtain a linear-time kernel for d-Set
Packing(k) as well.Comment: 18 page
Parallel Repetition of Entangled Games with Exponential Decay via the Superposed Information Cost
In a two-player game, two cooperating but non communicating players, Alice
and Bob, receive inputs taken from a probability distribution. Each of them
produces an output and they win the game if they satisfy some predicate on
their inputs/outputs. The entangled value of a game is the
maximum probability that Alice and Bob can win the game if they are allowed to
share an entangled state prior to receiving their inputs.
The -fold parallel repetition of consists of instances of
where the players receive all the inputs at the same time and produce all
the outputs at the same time. They win if they win each instance of .
In this paper we show that for any game such that , decreases exponentially in . First, for
any game on the uniform distribution, we show that , where and are the sizes of the input
and output sets. From this result, we show that for any entangled game ,
where is the input distribution of and
. This implies parallel
repetition with exponential decay as long as for
general games. To prove this parallel repetition, we introduce the concept of
\emph{Superposed Information Cost} for entangled games which is inspired from
the information cost used in communication complexity.Comment: In the first version of this paper we presented a different, stronger
Corollary 1 but due to an error in the proof we had to modify it in the
second version. This third version is a minor update. We correct some typos
and re-introduce a proof accidentally commented out in the second versio
On the Complexity of Computing Two Nonlinearity Measures
We study the computational complexity of two Boolean nonlinearity measures:
the nonlinearity and the multiplicative complexity. We show that if one-way
functions exist, no algorithm can compute the multiplicative complexity in time
given the truth table of length , in fact under the same
assumption it is impossible to approximate the multiplicative complexity within
a factor of . When given a circuit, the problem of
determining the multiplicative complexity is in the second level of the
polynomial hierarchy. For nonlinearity, we show that it is #P hard to compute
given a function represented by a circuit
On the Approximability of Digraph Ordering
Given an n-vertex digraph D = (V, A) the Max-k-Ordering problem is to compute
a labeling maximizing the number of forward edges, i.e.
edges (u,v) such that (u) < (v). For different values of k, this
reduces to Maximum Acyclic Subgraph (k=n), and Max-Dicut (k=2). This work
studies the approximability of Max-k-Ordering and its generalizations,
motivated by their applications to job scheduling with soft precedence
constraints. We give an LP rounding based 2-approximation algorithm for
Max-k-Ordering for any k={2,..., n}, improving on the known
2k/(k-1)-approximation obtained via random assignment. The tightness of this
rounding is shown by proving that for any k={2,..., n} and constant
, Max-k-Ordering has an LP integrality gap of 2 -
for rounds of the
Sherali-Adams hierarchy.
A further generalization of Max-k-Ordering is the restricted maximum acyclic
subgraph problem or RMAS, where each vertex v has a finite set of allowable
labels . We prove an LP rounding based
approximation for it, improving on the
approximation recently given by Grandoni et al.
(Information Processing Letters, Vol. 115(2), Pages 182-185, 2015). In fact,
our approximation algorithm also works for a general version where the
objective counts the edges which go forward by at least a positive offset
specific to each edge.
The minimization formulation of digraph ordering is DAG edge deletion or
DED(k), which requires deleting the minimum number of edges from an n-vertex
directed acyclic graph (DAG) to remove all paths of length k. We show that
both, the LP relaxation and a local ratio approach for DED(k) yield
k-approximation for any .Comment: 21 pages, Conference version to appear in ESA 201
Scheduling over Scenarios on Two Machines
We consider scheduling problems over scenarios where the goal is to find a
single assignment of the jobs to the machines which performs well over all
possible scenarios. Each scenario is a subset of jobs that must be executed in
that scenario and all scenarios are given explicitly. The two objectives that
we consider are minimizing the maximum makespan over all scenarios and
minimizing the sum of the makespans of all scenarios. For both versions, we
give several approximation algorithms and lower bounds on their
approximability. With this research into optimization problems over scenarios,
we have opened a new and rich field of interesting problems.Comment: To appear in COCOON 2014. The final publication is available at
link.springer.co
On the NP-Hardness of Approximating Ordering Constraint Satisfaction Problems
We show improved NP-hardness of approximating Ordering Constraint
Satisfaction Problems (OCSPs). For the two most well-studied OCSPs, Maximum
Acyclic Subgraph and Maximum Betweenness, we prove inapproximability of
and .
An OCSP is said to be approximation resistant if it is hard to approximate
better than taking a uniformly random ordering. We prove that the Maximum
Non-Betweenness Problem is approximation resistant and that there are width-
approximation-resistant OCSPs accepting only a fraction of
assignments. These results provide the first examples of
approximation-resistant OCSPs subject only to P \NP
Constant-degree graph expansions that preserve the treewidth
Many hard algorithmic problems dealing with graphs, circuits, formulas and
constraints admit polynomial-time upper bounds if the underlying graph has
small treewidth. The same problems often encourage reducing the maximal degree
of vertices to simplify theoretical arguments or address practical concerns.
Such degree reduction can be performed through a sequence of splittings of
vertices, resulting in an _expansion_ of the original graph. We observe that
the treewidth of a graph may increase dramatically if the splittings are not
performed carefully. In this context we address the following natural question:
is it possible to reduce the maximum degree to a constant without substantially
increasing the treewidth?
Our work answers the above question affirmatively. We prove that any simple
undirected graph G=(V, E) admits an expansion G'=(V', E') with the maximum
degree <= 3 and treewidth(G') <= treewidth(G)+1. Furthermore, such an expansion
will have no more than 2|E|+|V| vertices and 3|E| edges; it can be computed
efficiently from a tree-decomposition of G. We also construct a family of
examples for which the increase by 1 in treewidth cannot be avoided.Comment: 12 pages, 6 figures, the main result used by quant-ph/051107
- …